runtime.hmap.B (field)

67 uses

	runtime (current package)
		map.go#L122: 	B         uint8  // log_2 of # of buckets (can hold up to loadFactor * 2^B items)
		map.go#L232: 	if h.B < 16 {
		map.go#L239: 	mask := uint32(1)<<(h.B-15) - 1
		map.go#L323: 	h.B = B
		map.go#L328: 	if h.B != 0 {
		map.go#L330: 		h.buckets, nextOverflow = makeBucketArray(t, h.B, nil)
		map.go#L419: 	m := bucketMask(h.B)
		map.go#L480: 	m := bucketMask(h.B)
		map.go#L524: 	m := bucketMask(h.B)
		map.go#L609: 	bucket := hash & bucketMask(h.B)
		map.go#L658: 	if !h.growing() && (overLoadFactor(h.count+1, h.B) || tooManyOverflowBuckets(h.noverflow, h.B)) {
		map.go#L725: 	bucket := hash & bucketMask(h.B)
		map.go#L833: 	it.B = h.B
		map.go#L847: 	if h.B > 31-bucketCntBits {
		map.go#L852: 	it.startBucket = r & bucketMask(h.B)
		map.go#L853: 	it.offset = uint8(r >> h.B & (bucketCnt - 1))
		map.go#L890: 		if h.growing() && it.B == h.B {
		map.go#L1022: 	markBucketsEmpty(h.buckets, bucketMask(h.B))
		map.go#L1045: 	_, nextOverflow := makeBucketArray(t, h.B, h.buckets)
		map.go#L1063: 	if !overLoadFactor(h.count+1, h.B) {
		map.go#L1068: 	newbuckets, nextOverflow := makeBucketArray(t, h.B+bigger, nil)
		map.go#L1075: 	h.B += bigger
		map.go#L1133: 	oldB := h.B
		map.go#L1513: 	if src.B == 0 {
		map.go#L1521: 	if dst.B == 0 {
		map.go#L1524: 	dstArraySize := int(bucketShift(dst.B))
		map.go#L1525: 	srcArraySize := int(bucketShift(src.B))
		map.go#L1542: 	oldB := src.B
		map.go#L1555: 		if oldB >= dst.B { // main bucket bits in dst is less than oldB bits in src
		map.go#L1556: 			dstBmap := (*bmap)(add(dst.buckets, (uintptr(i)&bucketMask(dst.B))*uintptr(t.BucketSize)))
		map.go#L1610: 	offset := uint8(r >> h.B & (bucketCnt - 1))
		map.go#L1611: 	if h.B == 0 {
		map.go#L1615: 	arraySize := int(bucketShift(h.B))
		map.go#L1673: 	offset := uint8(r >> h.B & (bucketCnt - 1))
		map.go#L1674: 	if h.B == 0 {
		map.go#L1678: 	arraySize := int(bucketShift(h.B))
		map_fast32.go#L25: 	if h.B == 0 {
		map_fast32.go#L30: 		m := bucketMask(h.B)
		map_fast32.go#L65: 	if h.B == 0 {
		map_fast32.go#L70: 		m := bucketMask(h.B)
		map_fast32.go#L114: 	bucket := hash & bucketMask(h.B)
		map_fast32.go#L156: 	if !h.growing() && (overLoadFactor(h.count+1, h.B) || tooManyOverflowBuckets(h.noverflow, h.B)) {
		map_fast32.go#L204: 	bucket := hash & bucketMask(h.B)
		map_fast32.go#L246: 	if !h.growing() && (overLoadFactor(h.count+1, h.B) || tooManyOverflowBuckets(h.noverflow, h.B)) {
		map_fast32.go#L290: 	bucket := hash & bucketMask(h.B)
		map_fast64.go#L25: 	if h.B == 0 {
		map_fast64.go#L30: 		m := bucketMask(h.B)
		map_fast64.go#L65: 	if h.B == 0 {
		map_fast64.go#L70: 		m := bucketMask(h.B)
		map_fast64.go#L114: 	bucket := hash & bucketMask(h.B)
		map_fast64.go#L156: 	if !h.growing() && (overLoadFactor(h.count+1, h.B) || tooManyOverflowBuckets(h.noverflow, h.B)) {
		map_fast64.go#L204: 	bucket := hash & bucketMask(h.B)
		map_fast64.go#L246: 	if !h.growing() && (overLoadFactor(h.count+1, h.B) || tooManyOverflowBuckets(h.noverflow, h.B)) {
		map_fast64.go#L290: 	bucket := hash & bucketMask(h.B)
		map_faststr.go#L25: 	if h.B == 0 {
		map_faststr.go#L81: 	m := bucketMask(h.B)
		map_faststr.go#L120: 	if h.B == 0 {
		map_faststr.go#L176: 	m := bucketMask(h.B)
		map_faststr.go#L225: 	bucket := hash & bucketMask(h.B)
		map_faststr.go#L275: 	if !h.growing() && (overLoadFactor(h.count+1, h.B) || tooManyOverflowBuckets(h.noverflow, h.B)) {
		map_faststr.go#L319: 	bucket := hash & bucketMask(h.B)